Pipelined Training with Stale Weights in Deep Convolutional Neural Networks
نویسندگان
چکیده
The growth in size and complexity of convolutional neural networks (CNNs) is forcing the partitioning a network across multiple accelerators during training pipelining backpropagation computations over these accelerators. Pipelining results use stale weights. Existing approaches to pipelined avoid or limit weights with techniques that either underutilize increase memory footprint. This paper contributes scheme uses maximize accelerator utilization keep overhead modest. It explores impact on statistical efficiency performance using 4 CNNs (LeNet-5, AlexNet, VGG, ResNet) shows when introduced early layers, converges models comparable inference accuracies those resulting from nonpipelined (a drop accuracy 0.4%, 4%, 0.83%, 1.45% for networks, respectively). However, deeper network, significantly (up 12% VGG 8.5% ResNet-20). also hybrid combines address this drop. potential improvement proposed demonstrated proof-of-concept implementation PyTorch 2 GPUs ResNet-56/110/224/362, achieving speedups up 1.8X 1-GPU baseline.
منابع مشابه
Incremental Training of Deep Convolutional Neural Networks
We propose an incremental training method that partitions the original network into sub-networks, which are then gradually incorporated in the running network during the training process. To allow for a smooth dynamic growth of the network, we introduce a look-ahead initialization that outperforms the random initialization. We demonstrate that our incremental approach reaches the reference netw...
متن کاملBinaryConnect: Training Deep Neural Networks with binary weights during propagations
Deep Neural Networks (DNN) have achieved state-of-the-art results in a wide range of tasks, with the best results obtained with large training sets and large models. In the past, GPUs enabled these breakthroughs because of their greater computational speed. In the future, faster computation at both training and test time is likely to be crucial for further progress and for consumer applications...
متن کاملCystoscopy Image Classication Using Deep Convolutional Neural Networks
In the past three decades, the use of smart methods in medical diagnostic systems has attractedthe attention of many researchers. However, no smart activity has been provided in the eld ofmedical image processing for diagnosis of bladder cancer through cystoscopy images despite the highprevalence in the world. In this paper, two well-known convolutional neural networks (CNNs) ...
متن کاملTraining Deep Convolutional Neural Networks with Resistive Cross-Point Devices
In a previous work we have detailed the requirements for obtaining maximal deep learning performance benefit by implementing fully connected deep neural networks (DNN) in the form of arrays of resistive devices. Here we extend the concept of Resistive Processing Unit (RPU) devices to convolutional neural networks (CNNs). We show how to map the convolutional layers to fully connected RPU arrays ...
متن کاملTraining Deeper Convolutional Networks with Deep Supervision
One of the most promising ways of improving the performance of deep convolutional neural networks is by increasing the number of convolutional layers. However, adding layers makes training more difficult and computationally expensive. In order to train deeper networks, we propose to add auxiliary supervision branches after certain intermediate layers during training. We formulate a simple rule ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Applied Computational Intelligence and Soft Computing
سال: 2021
ISSN: ['1687-9724', '1687-9732']
DOI: https://doi.org/10.1155/2021/3839543